Stochastic Block BFGS: Squeezing More Curvature out of Data
نویسندگان
چکیده
is cheap (2), where Dt ∈ Rd×q and q min{d, n}. We employ three di erent sketching strategies: 1) gauss. Dt has standard Gaussian entries sampled i.i.d at each iteration. 2) prev. Let dt = −Htgt. Store search directions Dt = [dt+1−q , . . . , dt] and update Ht once every q iterations. 3) fact. Sample Ct ⊆ {1, . . . , d} uniformly at random and set Dt = Lt−1I:Ct ,where Lt−1L T t−1 = Ht−1 and I:Ct denotes the concatenation of the columns of the identity matrix indexed by a set Ct ⊂ {1, . . . , d}. 4. Block BFGS Update
منابع مشابه
The modified BFGS method with new secant relation for unconstrained optimization problems
Using Taylor's series we propose a modified secant relation to get a more accurate approximation of the second curvature of the objective function. Then, based on this modified secant relation we present a new BFGS method for solving unconstrained optimization problems. The proposed method make use of both gradient and function values while the usual secant relation uses only gradient values. U...
متن کاملGlobal convergence of online limited memory BFGS
Global convergence of an online (stochastic) limited memory version of the Broyden-FletcherGoldfarb-Shanno (BFGS) quasi-Newton method for solving optimization problems with stochastic objectives that arise in large scale machine learning is established. Lower and upper bounds on the Hessian eigenvalues of the sample functions are shown to suffice to guarantee that the curvature approximation ma...
متن کاملOn the Use of Stochastic Hessian Information in Optimization Methods for Machine Learning
This paper describes how to incorporate sampled curvature information in a NewtonCG method and in a limited memory quasi-Newton method for statistical learning. The motivation for this work stems from supervised machine learning applications involving a very large number of training points. We follow a batch approach, also known in the stochastic optimization literature as a sample average appr...
متن کاملBlock Bfgs Methods
We introduce a quasi-Newton method with block updates called Block BFGS. We show that this method, performed with inexact Armijo-Wolfe line searches, converges globally and superlinearly under the same convexity assumptions as BFGS. We also show that Block BFGS is globally convergent to a stationary point when applied to non-convex functions with bounded Hessian, and discuss other modifications...
متن کاملA Stochastic Quasi-Newton Method for Non-Rigid Image Registration
Image registration is often very slow because of the high dimensionality of the images and complexity of the algorithms. Adaptive stochastic gradient descent (ASGD) outperforms deterministic gradient descent and even quasi-Newton in terms of speed. This method, however, only exploits first-order information of the cost function. In this paper, we explore a stochastic quasi-Newton method (s-LBFG...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2016